swarm confrontation
Rule-Based Conflict-Free Decision Framework in Swarm Confrontation
Dong, Zhaoqi, Wang, Zhinan, Zheng, Quanqi, Xu, Bin, Chen, Lei, Lv, Jinhu
Traditional rule-based decision-making methods with interpretable advantage, such as finite state machine, suffer from the jitter or deadlock(JoD) problems in extremely dynamic scenarios. To realize agent swarm confrontation, decision conflicts causing many JoD problems are a key issue to be solved. Here, we propose a novel decision-making framework that integrates probabilistic finite state machine, deep convolutional networks, and reinforcement learning to implement interpretable intelligence into agents. Our framework overcomes state machine instability and JoD problems, ensuring reliable and adaptable decisions in swarm confrontation. The proposed approach demonstrates effective performance via enhanced human-like cooperation and competitive strategies in the rigorous evaluation of real experiments, outperforming other methods.
- North America > United States (0.14)
- Asia > China > Beijing > Beijing (0.05)
- South America > Argentina > Patagonia > Río Negro Province > Viedma (0.04)
- Leisure & Entertainment (0.93)
- Transportation (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.94)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Rule-Based Reasoning (0.71)
- (2 more...)
Hierarchical Reinforcement Learning for Swarm Confrontation with High Uncertainty
Wu, Qizhen, Liu, Kexin, Chen, Lei, Lü, Jinhu
In swarm robotics, confrontation including the pursuit-evasion game is a key scenario. High uncertainty caused by unknown opponents' strategies and dynamic obstacles complicates the action space into a hybrid decision process. Although the deep reinforcement learning method is significant for swarm confrontation since it can handle various sizes, as an end-to-end implementation, it cannot deal with the hybrid process. Here, we propose a novel hierarchical reinforcement learning approach consisting of a target allocation layer, a path planning layer, and the underlying dynamic interaction mechanism between the two layers, which indicates the quantified uncertainty. It decouples the hybrid process into discrete allocation and continuous planning layers, with a probabilistic ensemble model to quantify the uncertainty and regulate the interaction frequency adaptively. Furthermore, to overcome the unstable training process introduced by the two layers, we design an integration training method including pre-training and cross-training, which enhances the training efficiency and stability. Experiment results in both comparison and ablation studies validate the effectiveness and generalization performance of our proposed approach.
- Asia > China > Beijing > Beijing (0.05)
- Oceania > Australia > Victoria > Melbourne (0.04)
- North America > United States > New Jersey > Mercer County > Princeton (0.04)
- (4 more...)
- Government > Military (0.68)
- Leisure & Entertainment > Games (0.49)